Load Cache
نویسندگان
چکیده
As processors continue to exploit more instruction level parallelism, a greater demand is placed on reducing the eeects of memory access latency. In this paper, we introduce a novel modiication of the processor pipeline called memory renaming. Memory renaming applies register access techniques to load instructions, reducing the eeect of delays caused by the need to calculate effective addresses for the load and all preceding stores before the data can be fetched. Memory renaming allows the processor to speculatively fetch values when the producer of the data can be reliably determined without the need for an eeective address. This work extends previous studies of data value and dependence speculation. When memory renaming is added to the processor pipeline, renaming can be applied to 30% to 50% of all memory references, translating to an overall improvement in execution time of up to 41%. Furthermore, this improvement is seen across all memory segments { including the heap segment which has often been diicult to manage eeciently.
منابع مشابه
Achieving Load Balance and Effective Caching in Clustered Web Servers
This paper considers policies for distributing requests in clustered Web servers, wherein multiple server machines are configured to function as a single high(er) performance Web server. We evaluate various load distribution policies with respect to both their ability to achieve good load balance (the primary goal) and also to their impact on the effectiveness of per-machine caching. Trace-driv...
متن کاملAchieving Load Balance and Eeective Caching in Clustered Web Servers
This paper considers policies for distributing requests in clustered Web servers, wherein multiple server machines are conngured to function as a single high(er) performance Web server. We evaluate various load distribution policies with respect to both their ability to achieve good load balance (the primary goal) and also to their impact on the eeec-tiveness of per-machine caching. Trace-drive...
متن کاملEnabling Efficient Dynamic Resizing of Large DRAM Caches via A Hardware Consistent Hashing Mechanism
Die-stacked DRAM has been proposed for use as a large, high-bandwidth, last-level cache with hundreds or thousands of megabytes of capacity. Not all workloads (or phases) can productively utilize this much cache space, however. Unfortunately, the unused (or under-used) cache continues to consume power due to leakage in the peripheral circuitry and periodic DRAM refresh. Dynamically adjusting th...
متن کاملDomain Caching: Building Web Services for Live Events
Web servers providing live coverage of events are hard to design due to a high degree of variation in the load, variation in geographical distribution of users and the frequency with which content is updated. In this paper, we propose domain caching, a technique that dynamically distributes the load on the server by instructing selected clients (domain caches) to cache server content and servic...
متن کاملA Cache Design of Load Balancing System for Object-Based Storage
In design of the object-based storage system, cache design of the server becomes an important guarantee for improving service quality. This paper presents a cache design of load balancing system for object-based storage, which consists of three main aspects. First, the cache update policy is proposed. The strategy considers the overall cache as a pool, consistency and accuracy of updating is en...
متن کاملEC-Cache: Load-Balanced, Low-Latency Cluster Caching with Online Erasure Coding
Data-intensive clusters and object stores are increasingly relying on in-memory object caching to meet the I/O performance demands. These systems routinely face the challenges of popularity skew, background load imbalance, and server failures, which result in severe load imbalance across servers and degraded I/O performance. Selective replication is a commonly used technique to tackle these cha...
متن کامل